# Text Feature Extraction
Nomic Embed Text V2 Moe Unsupervised
This is an intermediate version of a multilingual Mixture of Experts (MoE) text embedding model, obtained through multi-stage contrastive training
Text Embedding
N
nomic-ai
161
5
Tybert
Apache-2.0
A Turkish BERT model pre-trained by Trendyol, suitable for various natural language understanding tasks.
Large Language Model
Transformers Other

T
Trendyol
54
6
Snowflake Arctic Embed M
Apache-2.0
Snowflake Arctic Embed M is a sentence transformer model focused on sentence similarity tasks, capable of efficiently extracting text features and calculating similarity between sentences.
Text Embedding
Transformers

S
Snowflake
722.08k
154
Labse En Ru Myv V2
This is a multilingual embedding model based on LaBSE, supporting Russian and Moksha languages, fine-tuned on the myv_ru_2022 dataset.
Large Language Model
Transformers Other

L
slone
59
0
Star
STAR is a supervised contrastive pretraining transformer model for social media writing style understanding.
Text Embedding
Transformers

S
AIDA-UPM
186
2
Simcse Ja Bert Base Clcmlp
This is a BERT-based Japanese SimCSE model, specifically designed for extracting high-quality sentence embeddings from Japanese sentences.
Text Embedding
Transformers Japanese

S
pkshatech
803
15
Sbert Chinese General V1
Apache-2.0
A general-purpose Chinese sentence embedding model for calculating sentence similarity and semantic search tasks.
Text Embedding
Transformers Chinese

S
DMetaSoul
388
6
Envibert
envibert is a bilingual model based on the RoBERTa architecture, supporting Vietnamese and English processing, optimized for production environments.
Large Language Model
Transformers Other

E
nguyenvulebinh
84
5
Debertav2 Base Uncased
Apache-2.0
BERT is a pre-trained language model based on the Transformer architecture, trained on English corpus through masked language modeling and next sentence prediction tasks.
Large Language Model English
D
mlcorelib
21
0
Chinese Roberta L 8 H 256
A Chinese RoBERTa model pretrained on CLUECorpusSmall, with 8 layers and 512 hidden units, suitable for various Chinese NLP tasks.
Large Language Model Chinese
C
uer
15
1
Distilbert Feature Extraction
DistilBERT is a lightweight distilled version of BERT, retaining 97% of BERT's performance while being 40% smaller in size.
Large Language Model
Transformers

D
julien-c
2,223
2
Medium Base
Apache-2.0
A Transformer model pre-trained on English corpus using self-supervised learning similar to ELECTRA, trained by predicting replaced tokens.
Large Language Model
Transformers English

M
funnel-transformer
69
0
Bert Base Irish Cased V1
gaBERT is a BERT-based monolingual Irish language model trained on 7.9 million Irish sentences, suitable for fine-tuning downstream tasks in Irish.
Large Language Model
Transformers

B
DCU-NLP
42
5
Featured Recommended AI Models